gradient descent - definizione. Che cos'è gradient descent
Diclib.com
Dizionario ChatGPT
Inserisci una parola o una frase in qualsiasi lingua 👆
Lingua:

Traduzione e analisi delle parole tramite l'intelligenza artificiale ChatGPT

In questa pagina puoi ottenere un'analisi dettagliata di una parola o frase, prodotta utilizzando la migliore tecnologia di intelligenza artificiale fino ad oggi:

  • come viene usata la parola
  • frequenza di utilizzo
  • è usato più spesso nel discorso orale o scritto
  • opzioni di traduzione delle parole
  • esempi di utilizzo (varie frasi con traduzione)
  • etimologia

Cosa (chi) è gradient descent - definizione


Gradient descent         
  • An animation showing the first 83 iterations of gradient descent applied to this example. Surfaces are [[isosurface]]s of <math>F(\mathbf{x}^{(n)})</math> at current guess <math>\mathbf{x}^{(n)}</math>, and arrows show the direction of descent. Due to a small and constant step size, the convergence is slow.
  • Illustration of gradient descent on a series of [[level set]]s
  • Fog in the mountains
  • The steepest descent algorithm applied to the [[Wiener filter]]<ref>Haykin, Simon S. Adaptive filter theory. Pearson Education India, 2008. - p. 108-142, 217-242</ref>
OPTIMIZATION ALGORITHM
Steepest descent; Gradient ascent; Gradient descent method; Steepest ascent; Gradient Descent; Gradient descent optimization; Gradient-based optimization; Gradient descent with momentum
In mathematics, gradient descent (also often called steepest descent) is a first-order iterative optimization algorithm for finding a local minimum of a differentiable function. The idea is to take repeated steps in the opposite direction of the gradient (or approximate gradient) of the function at the current point, because this is the direction of steepest descent.
Stochastic gradient descent         
OPTIMIZATION ALGORITHM
Gradient descent in machine learning; Incremental gradient descent; AdaGrad; Adagrad; Adam (optimization algorithm); Momentum (machine learning); RMSProp; Applications of stochastic gradient descent; Stochastic Gradient Descent; SGD optimizer; Adam optimizer; RMSprop
Stochastic gradient descent (often abbreviated SGD) is an iterative method for optimizing an objective function with suitable smoothness properties (e.g.
Gradient         
  • ''f''(''x'',''y'') {{=}} −(cos<sup>2</sup>''x'' + cos<sup>2</sup>''y'')<sup>2</sup>}} depicted as a projected [[vector field]] on the bottom plane.
  • 1=''f''(''x'', ''y'') = ''xe''<sup>−(''x''<sup>2</sup> + ''y''<sup>2</sup>)</sup>}} is plotted as arrows over the pseudocolor plot of the function.
MULTI-VARIABLE GENERALIZATION OF THE DERIVATIVE
Gradient vector; Gradients; Gradient (calculus); Gradient of a scalar; Gradient Operator; Grad operator
·adj Moving by steps; walking; as, gradient automata.
II. Gradient ·adj Adapted for walking, as the feet of certain birds.
III. Gradient ·noun The rate of regular or graded ascent or descent in a road; grade.
IV. Gradient ·noun A part of a road which slopes upward or downward; a portion of a way not level; a grade.
V. Gradient ·adj Rising or descending by regular degrees of inclination; as, the gradient line of a railroad.
VI. Gradient ·noun The rate of increase or decrease of a variable magnitude, or the curve which represents it; as, a thermometric gradient.